Fukumizu : Statistical Active Learning in Multilayer Perceptron 3

نویسنده

  • Kenji Fukumizu
چکیده

|This paper proposes new methods of generating input locations actively in gathering training data, aiming at solving problems special to multilayer perceptrons. One of the problems is that the optimum input locations which are calculated deterministically sometimes result in badly-distributed data and cause local minima in back-propagation training. Two probabilistic active learning methods, which utilize the statistical variance of locations, are proposed to solve this problem. One is parametric active learning and the other is multi-point-search active learning. Another serious problem in applying active learning to multilayer per-ceptrons is the singularity of a Fisher information matrix, whose regularity is assumed in many methods including the proposed ones. A technique of pruning redundant hidden units is proposed to keep the regularity of a Fisher information matrix, which makes active learning applicable to multilayer perceptrons. The eeectiveness of the proposed methods is demonstrated through computer simulations on simple artiicial problems and a real-world problem in color conversion.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Active Learning in Multilayer Perceptrons

We propose an active learning method with hidden-unit reduction, which is devised specially for multilayer perceptrons (MLP). First, we review our active learning method, and point out that many Fisher-information-based methods applied to MLP have a critical problem: the information matrix may be singular. To solve this problem, we derive the singularity condition of an information matrix, and ...

متن کامل

Dynamics of Batch Learning in Multilayer Neural Networks

We discuss the dynamics of batch learning of multilayer neural networks in the asymptotic limit, where the number of trining data is much larger than the number of parameters, emphasizing on the parameterization redundancy in overrealizable cases. In addition to showing experimental results on overtraining in multilayer perceptrons and three-layer linear neural networks, we theoretically prove ...

متن کامل

Effect of Batch Learning in Multilayerneural

This paper discusses batch gradient descent learning in mul-tilayer networks with a large number of statistical training data. We emphasize on the diierence between regular cases, where the prepared model has the same size as the true function , and overrealizable cases, where the model has surplus hidden units to realize the true function. First, experimental study on multilayer perceptrons an...

متن کامل

Asymptotic Theory of Locally Conic Models and its Application to Multilayer Neural Networks

This paper discusses the maximum likelihood estimation in a statistical model with unidentiÞability, using the framework of conic singularity. The likelihood ratio may diverge in unidentiÞable cases, though in regular cases it converges to a χ distribution. A useful sufficient condition of such divergence is obtained, and is applied to neural networks. The exact order for multilayer perceptrons...

متن کامل

Dynamics of Batch Learning in Multilayer

This paper investigates the dynamics of batch learning in multilayer neural networks. First, we present experimental results on the behavior in the steepest descent learning of multilayer perceptrons and linear neural networks. From the results of both models, we see that strong overtraining, the increase of generalization error, occurs in overrealizable cases where the target function is reali...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2000